Small-Vote Sample Selection for Label-Noise Learning

نویسندگان

چکیده

The small-loss criterion is widely used in recent label-noise learning methods. However, such a only considers the loss of each training sample mini-batch but ignores distribution whole set. Moreover, selection clean samples depends on heuristic data rate. As result, some noisy-labeled are easily identified as ones, and vice versa. In this paper, we propose novel yet simple method, which mainly consists Hierarchical Voting Scheme (HVS) an Adaptive Clean rate Estimation Strategy (ACES), to accurately identify for robust learning. Specifically, HVS effectively combine global vote local vote, so that both epoch-level batch-level information exploited assign hierarchical sample. Based HVS, further develop ACES adaptively estimate by leveraging 1D Gaussian Mixture Model (GMM). Experimental results show our proposed method consistently outperforms several state-of-the-art methods synthetic real-world noisy benchmark datasets.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

An Effective Approach for Robust Metric Learning in the Presence of Label Noise

Many algorithms in machine learning, pattern recognition, and data mining are based on a similarity/distance measure. For example, the kNN classifier and clustering algorithms such as k-means require a similarity/distance function. Also, in Content-Based Information Retrieval (CBIR) systems, we need to rank the retrieved objects based on the similarity to the query. As generic measures such as ...

متن کامل

Feature Selection for Small Sample Sets with High Dimensional Data Using Heuristic Hybrid Approach

Feature selection can significantly be decisive when analyzing high dimensional data, especially with a small number of samples. Feature extraction methods do not have decent performance in these conditions. With small sample sets and high dimensional data, exploring a large search space and learning from insufficient samples becomes extremely hard. As a result, neural networks and clustering a...

متن کامل

Feature Selection in Distance Learning from Small Sample

Learning from a small sample is an acute problem which arises in many applications where acquiring new samples is difficult, time consuming or expensive. The problem becomes even harder when dealing with rich high dimensional data. The learning process in such cases is often preceded by dimensionality reduction or feature selection. The need to avoid overfitting of an algorithm to the data is c...

متن کامل

Feature selection for multi-label learning

Feature Selection plays an important role in machine learning and data mining, and it is often applied as a data pre-processing step. This task can speed up learning algorithms and sometimes improve their performance. In multi-label learning, label dependence is considered another aspect that can contribute to improve learning performance. A replicable and wide systematic review performed by us...

متن کامل

Exploiting Multi-Label Information for Noise Resilient Feature Selection

In conventional supervised learning paradigm, each data instance is associated with one single class label. Multi-label learning differs in the way that data instances may belong to multiple concepts simultaneously, which naturally appear in a variety of high impact domains, ranging from bioinformatics, information retrieval to multimedia analysis. It targets to leverage the multiple label info...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Lecture Notes in Computer Science

سال: 2021

ISSN: ['1611-3349', '0302-9743']

DOI: https://doi.org/10.1007/978-3-030-86523-8_44